Domain Adaptation of a Dependency Parser with a Class-Class Selectional Preference Model
نویسندگان
چکیده
When porting parsers to a new domain, many of the errors are related to wrong attachment of out-of-vocabulary words. Since there is no available annotated data to learn the attachment preferences of the target domain words, we attack this problem using a model of selectional preferences based on domainspecific word classes. Our method uses Latent Dirichlet Allocations (LDA) to learn a domain-specific Selectional Preference model in the target domain using un-annotated data. The model provides features that model the affinities among pairs of words in the domain. To incorporate these new features in the parsing model, we adopt the co-training approach and retrain the parser with the selectional preferences features. We apply this method for adapting Easy First, a fast nondirectional parser trained on WSJ, to the biomedical domain (Genia Treebank). The Selectional Preference features reduce error by 4.5% over the co-training baseline.
منابع مشابه
Exploiting Web-Derived Selectional Preference to Improve Statistical Dependency Parsing
In this paper, we present a novel approach which incorporates the web-derived selectional preferences to improve statistical dependency parsing. Conventional selectional preference learning methods have usually focused on word-to-class relations, e.g., a verb selects as its subject a given nominal class. This paper extends previous work to wordto-word selectional preferences by using webscale d...
متن کاملHybrid Ways to Improve Domain Independence in an ML Dependency Parser
This paper reports a hybridization experi ment, where a baseline ML dependency pars er, LingPars, was allowed access to Con straint Grammar analyses provided by a rule based parser (EngGram) for the same data. Descriptive compatibility issues and their in fluence on performance are discussed. The hybrid system performed considerably better than its ML baseline, ...
متن کاملCmpt 825: Natural Language Processing 1.1 Hiding a Semantic Hierarchy in a Markov Model [1] 1.1.1 General Concepts
We know that in logic a predicate is a relation between its arguments. In other words, a predicate defines constraints between its arguments. A predicate ρ(v, r, c) is called selectional restriction where v is a verb, r is a role or an object and c is called a class which is a noun. Selectional preference σ : (v, r, c) → a is a function of these predicates to a real number. Where a shows the de...
متن کاملLearning class-to-class selectional preferences
Selectional preference learning methods have usually focused on wordto-class relations, e.g., a verb selects as its subject a given nominal class. This papers extends previous statistical models to class-to-class preferences, and presents a model that learns selectional preferences for classes of verbs. The motivation is twofold: different senses of a verb may have different preferences, and so...
متن کاملIntegrating selectional preferences in WordNet
Selectional preference learning methods have usually focused on word-to-class relations, e.g., a verb selects as its subject a given nominal class. This paper extends previous statistical models to class-to-class preferences, and presents a model that learns selectional preferences for classes of verbs, together with an algorithm to integrate the learned preferences in WordNet. The theoretical ...
متن کامل